<scp>Canine</scp>: Pre-training an Efficient Tokenization-Free Encoder for Language Representation

نویسندگان

چکیده

Pipelined NLP systems have largely been superseded by end-to-end neural modeling, yet nearly all commonly-used models still require an explicit tokenization step. While recent approaches based on data-derived subword lexicons are less brittle than manually engineered tokenizers, these techniques not equally suited to languages, and the use of any fixed vocabulary may limit a model's ability adapt. In this paper, we present CANINE, encoder that operates directly character sequences, without or vocabulary, pre-training strategy either characters optionally uses subwords as soft inductive bias. To its finer-grained input effectively efficiently, CANINE combines downsampling, which reduces sequence length, with deep transformer stack, encodes context. outperforms comparable mBERT model 2.8 F1 TyDi QA, challenging multilingual benchmark, despite having 28% fewer parameters.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

BPEmb: Tokenization-free Pre-trained Subword Embeddings in 275 Languages

We present BPEmb, a collection of pre-trained subword unit embeddings in 275 languages, based on Byte-Pair Encoding (BPE). In an evaluation using fine-grained entity typing as testbed, BPEmb performs competitively, and for some languages better than alternative subword approaches, while requiring vastly fewer resources and no tokenization. BPEmb is available at https://github.com/bheinzerling/b...

متن کامل

Multi-word tokenization for natural language processing

Sophisticated natural language processing (NLP) applications are entering everyday life in the form of translation services, electronic personal assistants or open-domain question answering systems. The more voice-operated applications like these become commonplace, the more expectations of users are raised to communicate with these services in unrestricted natural language, just as in a normal...

متن کامل

Auto-encoder pre-training of segmented-memory recurrent neural networks

The extended Backpropagation Through Time (eBPTT) learning algorithm for Segmented-Memory Recurrent Neural Networks (SMRNNs) yet lacks the ability to reliably learn long-term dependencies. The alternative learning algorithm, extended Real-Time Recurrent Learning (eRTRL), does not suffer this problem but is computational very intensive, such that it is impractical for the training of large netwo...

متن کامل

Efficient Subsampling for Training Complex Language Models

We propose an efficient way to train maximum entropy language models (MELM) and neural network language models (NNLM). The advantage of the proposed method comes from a more robust and efficient subsampling technique. The original multi-class language modeling problem is transformed into a set of binary problems where each binary classifier predicts whether or not a particular word will occur. ...

متن کامل

Efficient learning for spoken language understanding tasks with word embedding based pre-training

Spoken language understanding (SLU) tasks such as goal estimation and intention identification from user’s commands are essential components in spoken dialog systems. In recent years, neural network approaches have shown great success in various SLU tasks. However, one major difficulty of SLU is that the annotation of collected data can be expensive. Often this results in insufficient data bein...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Transactions of the Association for Computational Linguistics

سال: 2022

ISSN: ['2307-387X']

DOI: https://doi.org/10.1162/tacl_a_00448